2025-03-12 08:21:57.AIbase.16.2k
Alibaba's Tongyi Open-Sources R1-Omni Model for Enhanced Multimodal Emotion Recognition
On March 11th, the Tongyi Lab team announced the open-sourcing of the R1-Omni model, marking a significant advancement in multimodal model development. This model integrates reinforcement learning with verifiable reward (RLVR) methods, focusing on improving reasoning capabilities and generalization performance in multimodal emotion recognition tasks. R1-Omni's training is divided into two stages. In the cold-start phase, the team fine-tuned the model using a combined dataset containing 580 video clips sourced from Explainable Multimodal Emotion...